翻訳と辞書
Words near each other
・ Decipifus cingulatus
・ Decipium
・ Decipol
・ Decision
・ Decision (European Union)
・ Decision (novel)
・ Decision (TV series)
・ Decision 3012
・ Decision aids
・ Decision analysis
・ Decision analysis cycle
・ Decision analyst
・ Decision Analyst (company)
・ Decision at Sundown
・ Decision Before Dawn
Decision boundary
・ Decision cycle
・ Decision downloading
・ Decision Earth
・ Decision engineering
・ Decision EXpert
・ Decision fatigue
・ Decision field theory
・ Decision Game
・ Decision Games
・ Decision House
・ Decision in Paradise
・ Decision Lens
・ Decision Linear assumption
・ Decision list


Dictionary Lists
翻訳と辞書 辞書検索 [ 開発暫定版 ]
スポンサード リンク

Decision boundary : ウィキペディア英語版
Decision boundary

In a statistical-classification problem with two classes, a decision boundary or decision surface is a hypersurface that partitions the underlying vector space into two sets, one for each class. The classifier will classify all the points on one side of the decision boundary as belonging to one class and all those on the other side as belonging to the other class.
A decision boundary is the region of a problem space in which the output label of a classifier is ambiguous.〔http://www.cse.buffalo.edu/~jcorso/t/CSE555/files/quiz01_solutions.pdf〕
If the decision surface is a hyperplane, then the classification problem is linear, and the classes are linearly separable.
Decision boundaries are not always clear cut. That is, the transition from one class in the feature space to another is not discontinuous, but gradual. This effect is common in fuzzy logic based classification algorithms, where membership in one class or another is ambiguous.
==In ANNs and SVMs==
In the case of backpropagation based artificial neural networks or perceptrons, the type of decision boundary that the network can learn is determined by the number of hidden layers the network has.
If it has no hidden layers, then it can only learn linear problems. If it has one hidden layer, then it can learn problems with convex decision boundaries (and some concave decision boundaries). The network can learn more complex problems if it has two or more hidden layers.
In particular, support vector machines find a hyperplane that separates the feature space into two classes with the maximum margin. If the problem is not originally linearly separable, the kernel trick is used to turn it into a linearly separable one, by increasing the number of dimensions. Thus a general hypersurface in a small dimension space is turned into a hyperplane in a space with much larger dimensions.
Neural networks try to learn the decision boundary which minimizes the empirical error, while support vector machines try to learn the decision boundary which gives the best generalization.

抄文引用元・出典: フリー百科事典『 ウィキペディア(Wikipedia)
ウィキペディアで「Decision boundary」の詳細全文を読む



スポンサード リンク
翻訳と辞書 : 翻訳のためのインターネットリソース

Copyright(C) kotoba.ne.jp 1997-2016. All Rights Reserved.